Goto

Collaborating Authors

 lunarlander environment


Loss and Reward Weighing for increased learning in Distributed Reinforcement Learning

Holen, Martin, Andersen, Per-Arne, Knausgård, Kristian Muri, Goodwin, Morten

arXiv.org Artificial Intelligence

This paper introduces two learning schemes for distributed agents in Reinforcement Learning (RL) environments, namely Reward-Weighted (R-Weighted) and Loss-Weighted (L-Weighted) gradient merger. The R/L weighted methods replace standard practices for training multiple agents, such as summing or averaging the gradients. The core of our methods is to scale the gradient of each actor based on how high the reward (for R-Weighted) or the loss (for L-Weighted) is compared to the other actors. During training, each agent operates in differently initialized versions of the same environment, which gives different gradients from different actors. In essence, the R-Weights and L-Weights of each agent inform the other agents of its potential, which again reports which environment should be prioritized for learning. This approach of distributed learning is possible because environments that yield higher rewards, or low losses, have more critical information than environments that yield lower rewards or higher losses. We empirically demonstrate that the R-Weighted methods work superior to the state-of-the-art in multiple RL environments.


Reinforcement Learning(Part-1): Deep Q Learning using Tensorflow2

#artificialintelligence

In this tutorial, we will be discussing what is Q learning and how to Implement Q learning using Tensorflow2. Q Learning is one of the most popular reinforcement learning algorithms, it is an off-policy reinforcement learning(RL) that finds the best action for the given state. Q learning is considered off-policy reinforcement learning because Q learning is not dependent on current policy. It learns from actions that are outside the current policy, like taking random actions, therefore the policy is not needed. Here Q stands for Quality, which means how useful the given action is for the current state to get some future reward.


Policy Gradient(Reinforce)using Tensorflow2

#artificialintelligence

In this article, we will be discussing what is Policy gradients and how to implement policy gradients using tensorflow2. There are three main points in the policy gradient algorithm. By considering the above three principles, we can implement the policy gradient using TensorFlow. We are dividing our source code into two parts. Policy gradient takes the current state as input and outputs probabilities for all actions.


Dueling Double Deep Q Learning with Tensorflow

#artificialintelligence

In this article, we will be going through what is Dueling Double Deep Q Learning and how to implement it in Tenroflow. Dueling Double Deep Q learning is the combination of Dueling Deep Q Learning and Double Deep Q Learning. Let's try to understand what is Dueling Deep Q learning and Double Deep Q Learning. One of the drawbacks of the DQN algorithm is that it overestimates the true rewards; the Q-values think the agent is going to obtain a higher return than what it will obtain in reality. This overestimation is due to the presence of Max of Q value for the next state in the Q learning update equation.


Uncertainty-Based Out-of-Distribution Classification in Deep Reinforcement Learning

Sedlmeier, Andreas, Gabor, Thomas, Phan, Thomy, Belzner, Lenz, Linnhoff-Popien, Claudia

arXiv.org Machine Learning

Robustness to out-of-distribution (OOD) data is an important goal in building reliable machine learning systems. Especially in autonomous systems, wrong predictions for OOD inputs can cause safety critical situations. As a first step towards a solution, we consider the problem of detecting such data in a value-based deep reinforcement learning (RL) setting. Modelling this problem as a one-class classification problem, we propose a framework for uncertainty-based OOD classification: UBOOD. It is based on the effect that an agent's epistemic uncertainty is reduced for situations encountered during training (in-distribution), and thus lower than for unencountered (OOD) situations. Being agnostic towards the approach used for estimating epistemic uncertainty, combinations with different uncertainty estimation methods, e.g. approximate Bayesian inference methods or ensembling techniques are possible. We further present a first viable solution for calculating a dynamic classification threshold, based on the uncertainty distribution of the training data. Evaluation shows that the framework produces reliable classification results when combined with ensemble-based estimators, while the combination with concrete dropout-based estimators fails to reliably detect OOD situations. In summary, UBOOD presents a viable approach for OOD classification in deep RL settings by leveraging the epistemic uncertainty of the agent's value function.